71 research outputs found

    Generative Adversarial Positive-Unlabelled Learning

    Full text link
    In this work, we consider the task of classifying binary positive-unlabeled (PU) data. The existing discriminative learning based PU models attempt to seek an optimal reweighting strategy for U data, so that a decent decision boundary can be found. However, given limited P data, the conventional PU models tend to suffer from overfitting when adapted to very flexible deep neural networks. In contrast, we are the first to innovate a totally new paradigm to attack the binary PU task, from perspective of generative learning by leveraging the powerful generative adversarial networks (GAN). Our generative positive-unlabeled (GenPU) framework incorporates an array of discriminators and generators that are endowed with different roles in simultaneously producing positive and negative realistic samples. We provide theoretical analysis to justify that, at equilibrium, GenPU is capable of recovering both positive and negative data distributions. Moreover, we show GenPU is generalizable and closely related to the semi-supervised classification. Given rather limited P data, experiments on both synthetic and real-world dataset demonstrate the effectiveness of our proposed framework. With infinite realistic and diverse sample streams generated from GenPU, a very flexible classifier can then be trained using deep neural networks.Comment: 8 page

    Dynamic Feature Integration for Simultaneous Detection of Salient Object, Edge and Skeleton

    Full text link
    In this paper, we solve three low-level pixel-wise vision problems, including salient object segmentation, edge detection, and skeleton extraction, within a unified framework. We first show some similarities shared by these tasks and then demonstrate how they can be leveraged for developing a unified framework that can be trained end-to-end. In particular, we introduce a selective integration module that allows each task to dynamically choose features at different levels from the shared backbone based on its own characteristics. Furthermore, we design a task-adaptive attention module, aiming at intelligently allocating information for different tasks according to the image content priors. To evaluate the performance of our proposed network on these tasks, we conduct exhaustive experiments on multiple representative datasets. We will show that though these tasks are naturally quite different, our network can work well on all of them and even perform better than current single-purpose state-of-the-art methods. In addition, we also conduct adequate ablation analyses that provide a full understanding of the design principles of the proposed framework. To facilitate future research, source code will be released

    The difference between the domination number and the minus domination number of a cubic graph

    Get PDF
    AbstractThe closed neighborhood of a vertex subset S of a graph G = (V, E), denoted as N[S], is defined as the union of S and the set of all the vertices adjacent to some vertex of S. A dominating set of a graph G = (V, E) is defined as a set S of vertices such that N[S] = V. The domination number of a graph G, denoted as γ(G), is the minimum possible size of a dominating set of G. A minus dominating function on a graph G = (V, E) is a function g : V → {−1, 0, 1} such that g(N[v]) ≥ 1 for all vertices. The weight of a minus dominating function g is defined as g(V) =ΣvϵVg(v). The minus domination number of a graph G, denoted as γ−(G), is the minimum possible weight of a minus dominating function on G. It is well known that γ−(G) ≤ γ(G). This paper is focused on the difference between γ(G) and γ−(G) for cubic graphs. We first present a graph-theoretic description of γ−(G). Based on this, we give a necessary and sufficient condition for γ(G) −γ−(G) ≥ k. Further, we present an infinite family of cubic graphs of order 18k + 16 and with γ(G) −γ−(G) ≥

    S4Net: Single Stage Salient-Instance Segmentation

    Full text link
    We consider an interesting problem-salient instance segmentation in this paper. Other than producing bounding boxes, our network also outputs high-quality instance-level segments. Taking into account the category-independent property of each target, we design a single stage salient instance segmentation framework, with a novel segmentation branch. Our new branch regards not only local context inside each detection window but also its surrounding context, enabling us to distinguish the instances in the same scope even with obstruction. Our network is end-to-end trainable and runs at a fast speed (40 fps when processing an image with resolution 320x320). We evaluate our approach on a publicly available benchmark and show that it outperforms other alternative solutions. We also provide a thorough analysis of the design choices to help readers better understand the functions of each part of our network. The source code can be found at \url{https://github.com/RuochenFan/S4Net}
    • …
    corecore